Goto

Collaborating Authors

 meta model



Task-levelDifferentiallyPrivateMetaLearning

Neural Information Processing Systems

Specifically, meta learning takes in a collection of tasks (datasets) sampled from an unknown distribution. Each task defines a learning problem with respect to an input dataset.


Task-level Differentially Private Meta Learning

Neural Information Processing Systems

We study the problem of meta-learning with task-level differential privacy. Meta-learning has received increasing attention recently because of its ability to enable fast generalization to new task with small number of data points. However, the training process of meta learning likely involves exchange of task specific information, which may pose privacy risk especially in some privacy-sensitive applications. Therefore, it is important to provide strong privacy guarantees such that the learning process will not reveal any task sensitive information. To this end, existing works have proposed meta learning algorithms with record-level differential privacy, which is not sufficient in many scenarios since it does not protect the aggregated statistics based on the task dataset as a whole.



MFI-ResNet: Efficient ResNet Architecture Optimization via MeanFlow Compression and Selective Incubation

Sun, Nuolin, Wang, Linyuan, Wei, Haonan, Li, Lei, Yan, Bin

arXiv.org Artificial Intelligence

ResNet has achieved tremendous success in computer vision through its residual connection mechanism. ResNet can be viewed as a discretized form of ordinary differential equations (ODEs). From this perspective, the multiple residual blocks within a single ResNet stage essentially perform multi-step discrete iterations of the feature transformation for that stage. The recently proposed flow matching model, MeanFlow, enables one-step generative modeling by learning the mean velocity field to transform distributions. Inspired by this, we propose MeanFlow-Incubated ResNet (MFI-ResNet), which employs a compression-expansion strategy to jointly improve parameter efficiency and discriminative performance. In the compression phase, we simplify the multi-layer structure within each ResNet stage to one or two MeanFlow modules to construct a lightweight meta model. In the expansion phase, we apply a selective incubation strategy to the first three stages, expanding them to match the residual block configuration of the baseline ResNet model, while keeping the last stage in MeanFlow form, and fine-tune the incubated model. Experimental results show that on CIFAR-10 and CIFAR-100 datasets, MFI-ResNet achieves remarkable parameter efficiency, reducing parameters by 46.28% and 45.59% compared to ResNet-50, while still improving accuracy by 0.23% and 0.17%, respectively. This demonstrates that generative flow-fields can effectively characterize the feature transformation process in ResNet, providing a new perspective for understanding the relationship between generative modeling and discriminative learning.



A Proof of results in Section 4

Neural Information Processing Systems

A.1 Auxiliary lemmas Given a function f which is convex, L-Lipschitz and β -smooth. The proof mainly follows from Lemma 3.4 in [ Next we show the convergence rate of SGD with approximate gradients. By plugging into the value of T, η and b, we obtain the stated bound. In this section, we present the DP-FTMRL algorithm and the guarantee it provides.Algorithm 4: A More details of these functions can be found in Appendix B of [23]. Theorem 12. (Regret guarantee) Recall the settings in Theorem 10.



Distributionally robust minimization in meta-learning for system identification

Rufolo, Matteo, Piga, Dario, Forgione, Marco

arXiv.org Artificial Intelligence

-- Meta learning aims at learning how to solve tasks, and thus it allows to estimate models that can be quickly adapted to new scenarios. This work explores distributionally robust minimization in meta learning for system identification. Standard meta learning approaches optimize the expected loss, overlooking task variability. We use an alternative approach, adopting a distributionally robust optimization paradigm that prioritizes high-loss tasks, enhancing performance in worst-case scenarios. Evaluated on a meta model trained on a class of synthetic dynamical systems and tested in both in-distribution and out-of-distribution settings, the proposed approach allows to reduce failures in safety-critical applications. This often leads to suboptimal results when data is scarce. A promising alternative is to integrate meta learning, a framework introduced in the 1980s [1] and recently revitalized for its ability to enable rapid adaptation across related tasks. By training a meta-learner on a distribution of similar systems, the model can generalize efficiently to unseen dynamics with minimal additional data [2].


Task-level Differentially Private Meta Learning

Neural Information Processing Systems

We study the problem of meta-learning with task-level differential privacy. Meta-learning has received increasing attention recently because of its ability to enable fast generalization to new task with small number of data points. However, the training process of meta learning likely involves exchange of task specific information, which may pose privacy risk especially in some privacy-sensitive applications. Therefore, it is important to provide strong privacy guarantees such that the learning process will not reveal any task sensitive information. To this end, existing works have proposed meta learning algorithms with record-level differential privacy, which is not sufficient in many scenarios since it does not protect the aggregated statistics based on the task dataset as a whole.